Mutual Information, Relative Entropy and Estimation Error in Semi-Martingale Channels

نویسندگان
چکیده

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Estimation of Entropy and Mutual Information

We present some new results on the nonparametric estimation of entropy and mutual information. First, we use an exact local expansion of the entropy function to prove almost sure consistency and central limit theorems for three of the most commonly used discretized information estimators. The setup is related to Grenander’s method of sieves and places no assumptions on the underlying probabilit...

متن کامل

Entropy and Mutual Information for Markov Channels with General Inputs

We study new formulas based on Lyapunov exponents for entropy, mutual information, and capacity of finite state discrete time Markov channels. We also develop a method for directly computing mutual information and entropy using continuous state space Markov chains. Our methods allow for arbitrary input processes and channel dynamics, provided both have finite memory. We show that the entropy ra...

متن کامل

Geometric k-nearest neighbor estimation of entropy and mutual information

Nonparametric estimation of mutual information is used in a wide range of scientific problems to quantify dependence between variables. The k-nearest neighbor (knn) methods are consistent, and therefore expected to work well for a large sample size. These methods use geometrically regular local volume elements. This practice allows maximum localization of the volume elements, but can also induc...

متن کامل

Mimimal Relative Entropy Martingale Measure of Birth and Death Process

In this article, we investigate the MEMM (Minimal relative Entropy Martingale Measure) of Birth and Death processes and the MEMM of generalized Birth and Death processes. We see that the existence problem of the MEMM is reduced to the problem of solving the corresponding Hamilton-Jacobi-Bellman equation.

متن کامل

Efficient Entropy Estimation for Mutual Information Analysis Using B-Splines

The Correlation Power Analysis (CPA) is probably the most used side-channel attack because it seems to fit the power model of most standard CMOS devices and is very efficiently computed. However, the Pearson correlation coefficient used in the CPA measures only linear statistical dependences where the Mutual Information (MI) takes into account both linear and nonlinear dependences. Even if ther...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Information Theory

سال: 2018

ISSN: 0018-9448,1557-9654

DOI: 10.1109/tit.2018.2829889